Howard County
Appendix A V ariational Paragraph Embedder A.1 Selection of substitution rate p
Figure 4: Impact of the proportion of injected noise for learning Paragraph Em-beddings on XSum dataset. (Figure 4). The results of the ablation study are presented in Table 5. Embedder in providing clean and denoised reconstructions. In general, it has been observed that generations progress in a coarse-to-fine manner. The early time step, which is close to 1, tends to be less fluent and generic. This was the nicest stay we have ever had. Turtle Bay was a great resort. This was the nicest stay we have ever had.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- Oceania > Australia (0.04)
- North America > United States > Virginia (0.04)
- (12 more...)
- Oceania > Australia (0.04)
- South America > Colombia (0.04)
- Oceania > New Zealand (0.04)
- (3 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.94)
- Health & Medicine > Therapeutic Area > Dermatology (1.00)
- Health & Medicine > Therapeutic Area > Oncology (0.68)
Findings of the Fourth Shared Task on Multilingual Coreference Resolution: Can LLMs Dethrone Traditional Approaches?
Novák, Michal, Konopík, Miloslav, Nedoluzhko, Anna, Popel, Martin, Pražák, Ondřej, Sido, Jakub, Straka, Milan, Žabokrtský, Zdeněk, Zeman, Daniel
The paper presents an overview of the fourth edition of the Shared Task on Multilingual Coreference Resolution, organized as part of the CODI-CRAC 2025 workshop. As in the previous editions, participants were challenged to develop systems that identify mentions and cluster them according to identity coreference. A key innovation of this year's task was the introduction of a dedicated Large Language Model (LLM) track, featuring a simplified plaintext format designed to be more suitable for LLMs than the original CoNLL-U representation. The task also expanded its coverage with three new datasets in two additional languages, using version 1.3 of CorefUD - a harmonized multilingual collection of 22 datasets in 17 languages. In total, nine systems participated, including four LLM-based approaches (two fine-tuned and two using few-shot adaptation). While traditional systems still kept the lead, LLMs showed clear potential, suggesting they may soon challenge established approaches in future editions.
- Europe > Austria > Vienna (0.14)
- Europe > Hungary > Csongrád-Csanád County > Szeged (0.04)
- Europe > France > Provence-Alpes-Côte d'Azur > Bouches-du-Rhône > Marseille (0.04)
- (18 more...)
- Overview (0.86)
- Research Report (0.81)
Trusted Knowledge Extraction for Operations and Maintenance Intelligence
Mealey, Kathleen P., Karr, Jonathan A. Jr., Moreira, Priscila Saboia, Brenner, Paul R., Vardeman, Charles F. II
Deriving operational intelligence from organizational data repositories is a key challenge due to the dichotomy of data confidentiality vs data integration objectives, as well as the limitations of Natural Language Processing (NLP) tools relative to the specific knowledge structure of domains such as operations and maintenance. In this work, we discuss Knowledge Graph construction and break down the Knowledge Extraction process into its Named Entity Recognition, Coreference Resolution, Named Entity Linking, and Relation Extraction functional components. We then evaluate sixteen NLP tools in concert with or in comparison to the rapidly advancing capabilities of Large Language Models (LLMs). We focus on the operational and maintenance intelligence use case for trusted applications in the aircraft industry. A baseline dataset is derived from a rich public domain US Federal Aviation Administration dataset focused on equipment failures or maintenance requirements. We assess the zero-shot performance of NLP and LLM tools that can be operated within a controlled, confidential environment (no data is sent to third parties). Based on our observation of significant performance limitations, we discuss the challenges related to trusted NLP and LLM tools as well as their Technical Readiness Level for wider use in mission-critical industries such as aviation. We conclude with recommendations to enhance trust and provide our open-source curated dataset to support further baseline testing and evaluation.
- North America > United States > Maryland > Howard County > Columbia (0.04)
- South America > Colombia > Meta Department > Villavicencio (0.04)
- North America > United States > Virginia > Fairfax County > Fairfax (0.04)
- (15 more...)
- Transportation > Air (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
- Aerospace & Defense > Aircraft (1.00)
The Elephant in the Coreference Room: Resolving Coreference in Full-Length French Fiction Works
Bourgois, Antoine, Poibeau, Thierry
While coreference resolution is attracting more interest than ever from computational literature researchers, representative datasets of fully annotated long documents remain surprisingly scarce. In this paper, we introduce a new annotated corpus of three full-length French novels, totaling over 285,000 tokens. Unlike previous datasets focused on shorter texts, our corpus addresses the challenges posed by long, complex literary works, enabling evaluation of coreference models in the context of long reference chains. We present a modular coreference resolution pipeline that allows for fine-grained error analysis. We show that our approach is competitive and scales effectively to long documents. Finally, we demonstrate its usefulness to infer the gender of fictional characters, showcasing its relevance for both literary analysis and downstream NLP tasks.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > Austria > Vienna (0.14)
- North America > United States > Maryland > Howard County > Columbia (0.04)
- (31 more...)
Efficient Seq2seq Coreference Resolution Using Entity Representations
Grenander, Matt, Cohen, Shay B., Steedman, Mark
Seq2seq coreference models have introduced a new paradigm for coreference resolution by learning to generate text corresponding to coreference labels, without requiring task-specific parameters. While these models achieve new state-of-the-art performance, they do so at the cost of flexibility and efficiency. In particular, they do not efficiently handle incremental settings such as dialogue, where text must processed sequentially. We propose a compressed representation in order to improve the efficiency of these methods in incremental settings. Our method works by extracting and re-organizing entity-level tokens, and discarding the majority of other input tokens. On OntoNotes, our best model achieves just 0.6 CoNLL F1 points below a full-prefix, incremental baseline while achieving a compression ratio of 1.8. On LitBank, where singleton mentions are annotated, it passes state-of-the-art performance. Our results indicate that discarding a wide portion of tokens in seq2seq resolvers is a feasible strategy for incremental coreference resolution.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Asia > China > Hong Kong (0.05)
- North America > Dominican Republic (0.04)
- (17 more...)
ImCoref-CeS: An Improved Lightweight Pipeline for Coreference Resolution with LLM-based Checker-Splitter Refinement
Luo, Kangyang, Bai, Yuzhuo, Si, Shuzheng, Gao, Cheng, Wang, Zhitong, Shen, Yingli, Li, Wenhao, Liu, Zhu, Han, Yufeng, Wu, Jiayi, Kong, Cunliang, Sun, Maosong
Coreference Resolution (CR) is a critical task in Natural Language Processing (NLP). Current research faces a key dilemma: whether to further explore the potential of supervised neural methods based on small language models, whose detect-then-cluster pipeline still delivers top performance, or embrace the powerful capabilities of Large Language Models (LLMs). However, effectively combining their strengths remains underexplored. To this end, we propose \textbf{ImCoref-CeS}, a novel framework that integrates an enhanced supervised model with LLM-based reasoning. First, we present an improved CR method (\textbf{ImCoref}) to push the performance boundaries of the supervised neural method by introducing a lightweight bridging module to enhance long-text encoding capability, devising a biaffine scorer to comprehensively capture positional information, and invoking a hybrid mention regularization to improve training efficiency. Importantly, we employ an LLM acting as a multi-role Checker-Splitter agent to validate candidate mentions (filtering out invalid ones) and coreference results (splitting erroneous clusters) predicted by ImCoref. Extensive experiments demonstrate the effectiveness of ImCoref-CeS, which achieves superior performance compared to existing state-of-the-art (SOTA) methods.
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > Austria > Vienna (0.14)
- (15 more...)
- Law (1.00)
- Government > Regional Government > Africa Government (0.68)
- Banking & Finance (0.68)
- (2 more...)
Appendix A V ariational Paragraph Embedder A.1 Selection of substitution rate p
Figure 4: Impact of the proportion of injected noise for learning Paragraph Em-beddings on XSum dataset. (Figure 4). The results of the ablation study are presented in Table 5. Embedder in providing clean and denoised reconstructions. In general, it has been observed that generations progress in a coarse-to-fine manner. The early time step, which is close to 1, tends to be less fluent and generic. This was the nicest stay we have ever had. Turtle Bay was a great resort. This was the nicest stay we have ever had.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- Oceania > Australia (0.04)
- North America > United States > Virginia (0.04)
- (12 more...)
MR-UIE: Multi-Perspective Reasoning with Reinforcement Learning for Universal Information Extraction
Li, Zhongqiu, Wang, Shiquan, Fang, Ruiyu, Bao, Mengjiao, Wu, Zhenhe, Song, Shuangyong, Li, Yongxiang, He, Zhongjiang
Information extraction (IE) is a fundamental task in natural language processing (NLP), which encompasses a wide range of subtasks such as Named Entity Recognition (NER), Relation Extraction (RE), and Event Extraction (EE) [1-4]. Traditionally, these tasks have been addressed by specialized models trained in task-specific datasets. However, the fragmentation of tasks and schemas has hindered the development of generalizable and scalable IE tasks. To address this limitation, recent research has focused on universal information extraction (UIE), which aims to model all IE tasks within a universal framework. A seminal work in this direction is proposed by Lu et al., which introduced a structured generation paradigm that encodes diverse IE tasks into a common semantic representation[5]. Building on this, InstructUIE[6] extended the idea by incorporating multi-task instruction tuning, enabling models to generalize across tasks via natural language instructions. With the emergence of powerful LLMs[7-11], significant advancements have been made across long-standing NLP tasks such as text classification[12-16], intent recognition[17, 18], entity linking[19-22], and beyond. Inspired by their robust performance and adaptability, researchers have explored their potential for information extraction through prompting and in-context learning learning[23, 24]. For example, CodeIE demonstrated that code generation models can serve as strong few-shot IE extractors by using structured code-like commands[25].
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- Europe > Austria > Vienna (0.14)
- Asia > China > Beijing > Beijing (0.04)
- (10 more...)
TableZoomer: A Collaborative Agent Framework for Large-scale Table Question Answering
Xiong, Sishi, He, Ziyang, He, Zhongjiang, Zhao, Yu, Pan, Changzai, Zhang, Jie, Wu, Zhenhe, Song, Shuangyong, Li, Yongxiang
While large language models (LLMs) have shown promise in the table question answering (TQA) task through prompt engineering, they face challenges in industrial applications, including structural heterogeneity, difficulties in target data localization, and bottlenecks in complex reasoning. To address these limitations, this paper presents TableZoomer, a novel LLM-powered, programming-based agent framework. It introduces three key innovations: (1) replacing the original fully verbalized table with structured table schema to bridge the semantic gap and reduce computational complexity; (2) a query-aware table zooming mechanism that dynamically generates sub-table schema through column selection and entity linking, significantly improving target localization efficiency; and (3) a Program-of-Thoughts (PoT) strategy that transforms queries into executable code to mitigate numerical hallucination. Additionally, we integrate the reasoning workflow with the ReAct paradigm to enable iterative reasoning. Extensive experiments demonstrate that our framework maintains the usability advantages while substantially enhancing performance and scalability across tables of varying scales. When implemented with the Qwen3-8B-Instruct LLM, TableZoomer achieves accuracy improvements of 19.34% and 25% over conventional PoT methods on the large-scale DataBench dataset and the small-scale Fact Checking task of TableBench dataset, respectively.
- Europe > Austria > Vienna (0.14)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- North America > Mexico > Mexico City > Mexico City (0.04)
- (6 more...)